Brockton
Concept Bottleneck Large Language Models
Sun, Chung-En, Oikarinen, Tuomas, Ustun, Berk, Weng, Tsui-Wei
We introduce the Concept Bottleneck Large Language Model (CB-LLM), a pioneering approach to creating inherently interpretable Large Language Models (LLMs). Unlike traditional black-box LLMs that rely on post-hoc interpretation methods with limited neuron function insights, CB-LLM sets a new standard with its built-in interpretability, scalability, and ability to provide clear, accurate explanations. We investigate two essential tasks in the NLP domain: text classification and text generation. In text classification, CB-LLM narrows the performance gap with traditional black-box models and provides clear interpretability. In text generation, we show how interpretable neurons in CB-LLM can be used for concept detection and steering text generation. Our CB-LLMs enable greater interaction between humans and LLMs across a variety of tasks -- a feature notably absent in existing LLMs. Large Language Models (LLMs) have become instrumental in advancing Natural Language Processing (NLP) tasks.
- Asia > Middle East > Iraq > Baghdad Governorate > Baghdad (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Europe > Spain (0.04)
- (24 more...)
- Research Report > Promising Solution (0.34)
- Overview > Innovation (0.34)
- Media (1.00)
- Banking & Finance (1.00)
- Education (0.93)
- (3 more...)
Detecting Frames in News Headlines and Lead Images in U.S. Gun Violence Coverage
Tourni, Isidora Chara, Guo, Lei, Hu, Hengchang, Halim, Edward, Ishwar, Prakash, Daryanto, Taufiq, Jalal, Mona, Chen, Boqi, Betke, Margrit, Zhafransyah, Fabian, Lai, Sha, Wijaya, Derry Tanti
News media structure their reporting of events or issues using certain perspectives. When describing an incident involving gun violence, for example, some journalists may focus on mental health or gun regulation, while others may emphasize the discussion of gun rights. Such perspectives are called \say{frames} in communication research. We study, for the first time, the value of combining lead images and their contextual information with text to identify the frame of a given news article. We observe that using multiple modes of information(article- and image-derived features) improves prediction of news frames over any single mode of information when the images are relevant to the frames of the headlines. We also observe that frame image relevance is related to the ease of conveying frames via images, which we call frame concreteness. Additionally, we release the first multimodal news framing dataset related to gun violence in the U.S., curated and annotated by communication researchers. The dataset will allow researchers to further examine the use of multiple information modalities for studying media framing.
- North America > United States > Massachusetts > Plymouth County > Brockton (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > Canada > New Brunswick > York County > Fredericton (0.04)
- (15 more...)
- Media > News (1.00)
- Law > Criminal Law (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- (3 more...)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Vision (0.68)
- Information Technology > Sensing and Signal Processing > Image Processing (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
Learning Representations of Bi-level Knowledge Graphs for Reasoning beyond Link Prediction
Chung, Chanyoung, Whang, Joyce Jiyoung
Knowledge graphs represent known facts using triplets. While existing knowledge graph embedding methods only consider the connections between entities, we propose considering the relationships between triplets. For example, let us consider two triplets $T_1$ and $T_2$ where $T_1$ is (Academy_Awards, Nominates, Avatar) and $T_2$ is (Avatar, Wins, Academy_Awards). Given these two base-level triplets, we see that $T_1$ is a prerequisite for $T_2$. In this paper, we define a higher-level triplet to represent a relationship between triplets, e.g., $\langle T_1$, PrerequisiteFor, $T_2\rangle$ where PrerequisiteFor is a higher-level relation. We define a bi-level knowledge graph that consists of the base-level and the higher-level triplets. We also propose a data augmentation strategy based on the random walks on the bi-level knowledge graph to augment plausible triplets. Our model called BiVE learns embeddings by taking into account the structures of the base-level and the higher-level triplets, with additional consideration of the augmented triplets. We propose two new tasks: triplet prediction and conditional link prediction. Given a triplet $T_1$ and a higher-level relation, the triplet prediction predicts a triplet that is likely to be connected to $T_1$ by the higher-level relation, e.g., $\langle T_1$, PrerequisiteFor, ?$\rangle$. The conditional link prediction predicts a missing entity in a triplet conditioned on another triplet, e.g., $\langle T_1$, PrerequisiteFor, (Avatar, Wins, ?)$\rangle$. Experimental results show that BiVE significantly outperforms all other methods in the two new tasks and the typical base-level link prediction in real-world bi-level knowledge graphs.
- North America > United States > New York (0.05)
- Europe > Sweden > Stockholm > Stockholm (0.04)
- Europe > Italy (0.04)
- (12 more...)
- Research Report > New Finding (0.34)
- Personal > Honors (0.34)
- Media > Music (1.00)
- Media > Film (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- (2 more...)
Controlling Hallucinations at Word Level in Data-to-Text Generation
Rebuffel, Clément, Roberti, Marco, Soulier, Laure, Scoutheeten, Geoffrey, Cancelliere, Rossella, Gallinari, Patrick
Data-to-Text Generation (DTG) is a subfield of Natural Language Generation aiming at transcribing structured data in natural language descriptions. The field has been recently boosted by the use of neural-based generators which exhibit on one side great syntactic skills without the need of hand-crafted pipelines; on the other side, the quality of the generated text reflects the quality of the training data, which in realistic settings only offer imperfectly aligned structure-text pairs. Consequently, state-of-art neural models include misleading statements - usually called hallucinations - in their outputs. The control of this phenomenon is today a major challenge for DTG, and is the problem addressed in the paper. Previous work deal with this issue at the instance level: using an alignment score for each table-reference pair. In contrast, we propose a finer-grained approach, arguing that hallucinations should rather be treated at the word level. Specifically, we propose a Multi-Branch Decoder which is able to leverage word-level labels to learn the relevant parts of each training instance. These labels are obtained following a simple and efficient scoring procedure based on co-occurrence analysis and dependency parsing. Extensive evaluations, via automated metrics and human judgment on the standard WikiBio benchmark, show the accuracy of our alignment labels and the effectiveness of the proposed Multi-Branch Decoder. Our model is able to reduce and control hallucinations, while keeping fluency and coherence in generated texts. Further experiments on a degraded version of ToTTo show that our model could be successfully used on very noisy settings.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Europe > Germany > Schleswig-Holstein (0.05)
- Asia > Maldives (0.05)
- (17 more...)
- Media (1.00)
- Leisure & Entertainment > Sports (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Inductive Learning (0.87)
- Information Technology > Artificial Intelligence > Natural Language > Machine Translation (0.68)
- Information Technology > Artificial Intelligence > Natural Language > Grammars & Parsing (0.66)
Robot kayaks found the basin of an Alaskan glacier is melting 100 TIMES faster than models showed
Seaborne robots have made a startling discovery beneath a 20-mile glacier in Alaska. The technology found the massive rivers of ice may be melting under the LeConte Glacier much faster than previously thought. Scientists programmed autonomous kayaks to swim near the icy cliffs of the glacier to measure the'ambient meltwater intrusions', which shows how much fresh water is flowing into the ocean from underneath the glacier. The study found ambient melting was 100 times higher than models had estimated. This is the first time experts have been able to analyze plumes of meltwater - the water released when snow or ice melts, where glaciers meet the ocean- because the feat is far too dangerous for ships due to falling ice of slabs from the glacier.
- North America > Canada (0.14)
- North America > Greenland (0.05)
- Africa > Senegal (0.05)
- (18 more...)
Global Big Data Conference
Until last year, Brandon Tory, a senior software Artificial Intelligence engineer at Google and rapper, led a secret double life. Raised in Brockton, Massachusetts, a neighborhood known for crime and drugs, he lived with his family in a shelter as a teenager. Tory knew he wanted to be some kind of scientist, but also had a passion for music. He wanted to have a huge impact in both creativity and science. At University of Massachusetts Amherst he studied computer engineering and then worked as a senior Apple engineer in Cupertino.
- North America > United States > Massachusetts > Plymouth County > Brockton (0.26)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.26)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
- Information Technology > Data Science > Data Mining > Big Data (0.40)
- Information Technology > Artificial Intelligence (0.37)
AI for Crime Prevention and Detection - 5 Current Applications
Companies and cities all over world are experimenting with using artificial intelligence to reduce and prevent crime, and to more quickly respond to crimes in progress. The ideas behind many of these projects is that crimes are relatively predictable; it just requires being able to sort through a massive volume of data to find patterns that are useful to law enforcement. This kind of data analysis was technologically impossible a few decades ago, but the hope is that recent developments in machine learning are up to the task. There is good reason why companies and government are both interested in trying to use AI in this manner. As of 2010, the United States spent over $80 billion a year on incarations at the state, local, and federal levels. Estimates put the United States' total spending on law enforcement at over $100 billion a year. Law enforcement and prisons make up a substantial percentage of local government budgets.
- North America > United States > Wisconsin (0.04)
- North America > United States > Washington > Pierce County > Tacoma (0.04)
- North America > United States > Pennsylvania (0.04)
- (16 more...)